51 research outputs found

    Smoothed Complexity Theory

    Get PDF
    Smoothed analysis is a new way of analyzing algorithms introduced by Spielman and Teng (J. ACM, 2004). Classical methods like worst-case or average-case analysis have accompanying complexity classes, like P and AvgP, respectively. While worst-case or average-case analysis give us a means to talk about the running time of a particular algorithm, complexity classes allows us to talk about the inherent difficulty of problems. Smoothed analysis is a hybrid of worst-case and average-case analysis and compensates some of their drawbacks. Despite its success for the analysis of single algorithms and problems, there is no embedding of smoothed analysis into computational complexity theory, which is necessary to classify problems according to their intrinsic difficulty. We propose a framework for smoothed complexity theory, define the relevant classes, and prove some first hardness results (of bounded halting and tiling) and tractability results (binary optimization problems, graph coloring, satisfiability). Furthermore, we discuss extensions and shortcomings of our model and relate it to semi-random models.Comment: to be presented at MFCS 201

    Computing Diffusion State Distance using Green's Function and Heat Kernel on Graphs

    Full text link
    The diffusion state distance (DSD) was introduced by Cao-Zhang-Park-Daniels-Crovella-Cowen-Hescott [{\em PLoS ONE, 2013}] to capture functional similarity in protein-protein interaction networks. They proved the convergence of DSD for non-bipartite graphs. In this paper, we extend the DSD to bipartite graphs using lazy-random walks and consider the general LqL_q-version of DSD. We discovered the connection between the DSD LqL_q-distance and Green's function, which was studied by Chung and Yau [{\em J. Combinatorial Theory (A), 2000}]. Based on that, we computed the DSD LqL_q-distance for Paths, Cycles, Hypercubes, as well as random graphs G(n,p)G(n,p) and G(w1,...,wn)G(w_1,..., w_n). We also examined the DSD distances of two biological networks.Comment: Accepted by the 11th Workshop on Algorithms and Models for the Web Graph (WAW2014

    On the cavity method for decimated random constraint satisfaction problems and the analysis of belief propagation guided decimation algorithms

    Full text link
    We introduce a version of the cavity method for diluted mean-field spin models that allows the computation of thermodynamic quantities similar to the Franz-Parisi quenched potential in sparse random graph models. This method is developed in the particular case of partially decimated random constraint satisfaction problems. This allows to develop a theoretical understanding of a class of algorithms for solving constraint satisfaction problems, in which elementary degrees of freedom are sequentially assigned according to the results of a message passing procedure (belief-propagation). We confront this theoretical analysis to the results of extensive numerical simulations.Comment: 32 pages, 24 figure

    Sparser Random 3SAT Refutation Algorithms and the Interpolation Problem:Extended Abstract

    Get PDF
    We formalize a combinatorial principle, called the 3XOR principle, due to Feige, Kim and Ofek [12], as a family of unsatisfiable propositional formulas for which refutations of small size in any propo-sitional proof system that possesses the feasible interpolation property imply an efficient deterministic refutation algorithm for random 3SAT with n variables and ℩(n1.4) clauses. Such small size refutations would improve the state of the art (with respect to the clause density) efficient refutation algorithm, which works only for ℩(n1.5) many clauses [13]. We demonstrate polynomial-size refutations of the 3XOR principle in resolution operating with disjunctions of quadratic equations with small integer coefficients, denoted R(quad); this is a weak extension of cutting planes with small coefficients. We show that R(quad) is weakly autom-atizable iff R(lin) is weakly automatizable, where R(lin) is similar to R(quad) but with linear instead of quadratic equations (introduced in [25]). This reduces the problem of refuting random 3CNF with n vari-ables and ℩(n1.4) clauses to the interpolation problem of R(quad) and to the weak automatizability of R(lin)

    On Eigenvalues of Random Complexes

    Full text link
    We consider higher-dimensional generalizations of the normalized Laplacian and the adjacency matrix of graphs and study their eigenvalues for the Linial-Meshulam model Xk(n,p)X^k(n,p) of random kk-dimensional simplicial complexes on nn vertices. We show that for p=Ω(log⁥n/n)p=\Omega(\log n/n), the eigenvalues of these matrices are a.a.s. concentrated around two values. The main tool, which goes back to the work of Garland, are arguments that relate the eigenvalues of these matrices to those of graphs that arise as links of (k−2)(k-2)-dimensional faces. Garland's result concerns the Laplacian; we develop an analogous result for the adjacency matrix. The same arguments apply to other models of random complexes which allow for dependencies between the choices of kk-dimensional simplices. In the second part of the paper, we apply this to the question of possible higher-dimensional analogues of the discrete Cheeger inequality, which in the classical case of graphs relates the eigenvalues of a graph and its edge expansion. It is very natural to ask whether this generalizes to higher dimensions and, in particular, whether the higher-dimensional Laplacian spectra capture the notion of coboundary expansion - a generalization of edge expansion that arose in recent work of Linial and Meshulam and of Gromov. We show that this most straightforward version of a higher-dimensional discrete Cheeger inequality fails, in quite a strong way: For every k≄2k\geq 2 and n∈Nn\in \mathbb{N}, there is a kk-dimensional complex YnkY^k_n on nn vertices that has strong spectral expansion properties (all nontrivial eigenvalues of the normalised kk-dimensional Laplacian lie in the interval [1−O(1/n),1+O(1/n)][1-O(1/\sqrt{n}),1+O(1/\sqrt{n})]) but whose coboundary expansion is bounded from above by O(log⁥n/n)O(\log n/n) and so tends to zero as n→∞n\rightarrow \infty; moreover, YnkY^k_n can be taken to have vanishing integer homology in dimension less than kk.Comment: Extended full version of an extended abstract that appeared at SoCG 2012, to appear in Israel Journal of Mathematic

    Analysis of an exhaustive search algorithm in random graphs and the n^{c\log n} -asymptotics

    Full text link
    We analyze the cost used by a naive exhaustive search algorithm for finding a maximum independent set in random graphs under the usual G_{n,p} -model where each possible edge appears independently with the same probability p. The expected cost turns out to be of the less common asymptotic order n^{c\log n}, which we explore from several different perspectives. Also we collect many instances where such an order appears, from algorithmics to analysis, from probability to algebra. The limiting distribution of the cost required by the algorithm under a purely idealized random model is proved to be normal. The approach we develop is of some generality and is amenable for other graph algorithms.Comment: 35 page

    Finding planted partitions in random graphs with general degree distributions

    Get PDF
    We consider the problem of recovering a planted partition such as a coloring, a small bisection, or a large cut in an (apart from that) random graph. In the last 30 years many algorithms for this problem have been developed that work provably well on various random graph models resembling the ErdƑs–RĂ©nyi model Gn,m. In these random graph models edges are distributed uniformly, and thus the degree distribution is very regular. By contrast, the recent theory of large networks shows that real-world networks frequently have a significantly different distribution of the edges and hence also a different degree distribution. Therefore, a variety of new types of random graphs have been introduced to capture these specific properties. One of the most popular models is characterized by a prescribed expected degree sequence. We study a natural variant of this model that features a planted partition. Our main result is that there is a polynomial time algorithm for recovering (a large part of) the planted partition in this model even in the sparse case, where the average degree is constant. In contrast to prior work, the input of the algorithm consists only of the graph, i.e., no further parameters of the model (such as the expected degree sequence) are revealed to the algorithm
    • 

    corecore